Continuity of Optimal Values and Solutions for Control of Markov Chains with Constraints
نویسندگان
چکیده
We consider in this paper constrained Markov decision processes. This type of control model has many applications in telecommunications and other fields [E. Altman and A. Shwartz, IEEE Trans. Automat. Control, 34 (1989), pp. 1089–1102, E. A. Feinberg and M. I. Reiman, Probab. Engrg. Inform. Sci., 8 (1994), pp. 463–489, A. Hordijk and F. Spieksma, Adv. in Appl. Probab., 21 (1989), pp. 409–431, A. Lazar, IEEE Trans. Automat. Control, 28 (1983), pp. 1001–1007, P. Nain and K. W. Ross, IEEE Trans. Automat. Control, 31 (1986), pp. 883–888, K. W. Ross and B. Chen, IEEE Trans. Automat. Control, 33 (1988), pp. 261–267]. We address the issue of the convergence of the value and optimal policies of the problem with discounted costs, to the ones for the problem with expected average cost. We consider the general multichain ergodic structure. We present two stability results in this paper. We establish the continuity of optimal values and solutions of as well as some type of robustness of some suboptimal solutions in the discount factor. Our proof relies on same general theory on continuity of values and solutions in convex optimization that relies on well-known notions of Γ-convergence.
منابع مشابه
Stochastic Dynamic Programming with Markov Chains for Optimal Sustainable Control of the Forest Sector with Continuous Cover Forestry
We present a stochastic dynamic programming approach with Markov chains for optimal control of the forest sector. The forest is managed via continuous cover forestry and the complete system is sustainable. Forest industry production, logistic solutions and harvest levels are optimized based on the sequentially revealed states of the markets. Adaptive full system optimization is necessary for co...
متن کاملRobustness of convex optimization with application to controled Markov chains
We present two stability results in this paper. We rst obtain suucient conn ditions for the continuity of optimal values and solutions of convex programs in general vector spaces, as well as some types of robustness of some sub-optimal solutions. We then use these results in order to establish a new result in stochastic dynamic control of discrete event systems (known as constrained Markov Deci...
متن کاملA new machine replacement policy based on number of defective items and Markov chains
A novel optimal single machine replacement policy using a single as well as a two-stage decision making process is proposed based on the quality of items produced. In a stage of this policy, if the number of defective items in a sample of produced items is more than an upper threshold, the machine is replaced. However, the machine is not replaced if the number of defective items is less than ...
متن کاملFinancial Risk Modeling with Markova Chain
Investors use different approaches to select optimal portfolio. so, Optimal investment choices according to return can be interpreted in different models. The traditional approach to allocate portfolio selection called a mean - variance explains. Another approach is Markov chain. Markov chain is a random process without memory. This means that the conditional probability distribution of the nex...
متن کاملA Neural Network Method Based on Mittag-Leffler Function for Solving a Class of Fractional Optimal Control Problems
In this paper, a computational intelligence method is used for the solution of fractional optimal control problems (FOCP)'s with equality and inequality constraints. According to the Ponteryagin minimum principle (PMP) for FOCP with fractional derivative in the Riemann- Liouville sense and by constructing a suitable error function, we define an unconstrained minimization problem. In the optimiz...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- SIAM J. Control and Optimization
دوره 38 شماره
صفحات -
تاریخ انتشار 2000